Goto

Collaborating Authors

 resistance distance




Efficient Exact Resistance Distance Computation on Small-Treewidth Graphs: a Labelling Approach

Liao, Meihao, Pan, Yueyang, Li, Rong-Hua, Wang, Guoren

arXiv.org Artificial Intelligence

Resistance distance computation is a fundamental problem in graph analysis, yet existing random walk-based methods are limited to approximate solutions and suffer from poor efficiency on small-treewidth graphs (e.g., road networks). In contrast, shortest-path distance computation achieves remarkable efficiency on such graphs by leveraging cut properties and tree decompositions. Motivated by this disparity, we first analyze the cut property of resistance distance. While a direct generalization proves impractical due to costly matrix operations, we overcome this limitation by integrating tree decompositions, revealing that the resistance distance $r(s,t)$ depends only on labels along the paths from $s$ and $t$ to the root of the decomposition. This insight enables compact labelling structures. Based on this, we propose \treeindex, a novel index method that constructs a resistance distance labelling of size $O(n \cdot h_{\mathcal{G}})$ in $O(n \cdot h_{\mathcal{G}}^2 \cdot d_{\max})$ time, where $h_{\mathcal{G}}$ (tree height) and $d_{\max}$ (maximum degree) behave as small constants in many real-world small-treewidth graphs (e.g., road networks). Our labelling supports exact single-pair queries in $O(h_{\mathcal{G}})$ time and single-source queries in $O(n \cdot h_{\mathcal{G}})$ time. Extensive experiments show that TreeIndex substantially outperforms state-of-the-art approaches. For instance, on the full USA road network, it constructs a $405$ GB labelling in $7$ hours (single-threaded) and answers exact single-pair queries in $10^{-3}$ seconds and single-source queries in $190$ seconds--the first exact method scalable to such large graphs.


SALMAN: Stability Analysis of Language Models Through the Maps Between Graph-based Manifolds

Cheng, Wuxinlin, Cao, Yupeng, Wu, Jinwen, Subbalakshmi, Koduvayur, Han, Tian, Feng, Zhuo

arXiv.org Artificial Intelligence

Recent strides in pretrained transformer-based language models have propelled state-of-the-art performance in numerous NLP tasks. Yet, as these models grow in size and deployment, their robustness under input perturbations becomes an increasingly urgent question. Existing robustness methods often diverge between small-parameter and large-scale models (LLMs), and they typically rely on labor-intensive, sample-specific adversarial designs. In this paper, we propose a unified, local (sample-level) robustness framework (SALMAN) that evaluates model stability without modifying internal parameters or resorting to complex perturbation heuristics. Central to our approach is a novel Distance Mapping Distortion (DMD) measure, which ranks each sample's susceptibility by comparing input-to-output distance mappings in a near-linear complexity manner. By demonstrating significant gains in attack efficiency and robust training, we position our framework as a practical, model-agnostic tool for advancing the reliability of transformer-based NLP systems.


On Feature Decorrelation in Cloth-Changing Person Re-identification

Wang, Hongjun, Chen, Jiyuan, Jiang, Renhe, Song, Xuan, Zheng, Yinqiang

arXiv.org Artificial Intelligence

Cloth-changing person re-identification (CC-ReID) poses a significant challenge in computer vision. A prevailing approach is to prompt models to concentrate on causal attributes, like facial features and hairstyles, rather than confounding elements such as clothing appearance. Traditional methods to achieve this involve integrating multi-modality data or employing manually annotated clothing labels, which tend to complicate the model and require extensive human effort. In our study, we demonstrate that simply reducing feature correlations during training can significantly enhance the baseline model's performance. We theoretically elucidate this effect and introduce a novel regularization technique based on density ratio estimation. This technique aims to minimize feature correlation in the training process of cloth-changing ReID baselines. Our approach is model-independent, offering broad enhancements without needing additional data or labels. We validate our method through comprehensive experiments on prevalent CC-ReID datasets, showing its effectiveness in improving baseline models' generalization capabilities.


When does the mean network capture the topology of a sample of networks?

Meyer, François G

arXiv.org Machine Learning

The notion of Fr\'echet mean (also known as "barycenter") network is the workhorse of most machine learning algorithms that require the estimation of a "location" parameter to analyse network-valued data. In this context, it is critical that the network barycenter inherits the topological structure of the networks in the training dataset. The metric - which measures the proximity between networks - controls the structural properties of the barycenter. This work is significant because it provides for the first time analytical estimates of the sample Fr\'echet mean for the stochastic blockmodel, which is at the cutting edge of rigorous probabilistic analysis of random networks. We show that the mean network computed with the Hamming distance is unable to capture the topology of the networks in the training sample, whereas the mean network computed using the effective resistance distance recovers the correct partitions and associated edge density. From a practical standpoint, our work informs the choice of metrics in the context where the sample Fr\'echet mean network is used to characterise the topology of networks for network-valued machine learning


Comparing Graph Transformers via Positional Encodings

Black, Mitchell, Wan, Zhengchao, Mishne, Gal, Nayyeri, Amir, Wang, Yusu

arXiv.org Artificial Intelligence

The distinguishing power of graph transformers is closely tied to the choice of positional encoding: features used to augment the base transformer with information about the graph. There are two primary types of positional encoding: absolute positional encodings (APEs) and relative positional encodings (RPEs). APEs assign features to each node and are given as input to the transformer. RPEs instead assign a feature to each pair of nodes, e.g., graph distance, and are used to augment the attention block. A priori, it is unclear which method is better for maximizing the power of the resulting graph transformer. In this paper, we aim to understand the relationship between these different types of positional encodings. Interestingly, we show that graph transformers using APEs and RPEs are equivalent in terms of distinguishing power. In particular, we demonstrate how to interchange APEs and RPEs while maintaining their distinguishing power in terms of graph transformers. Based on our theoretical results, we provide a study on several APEs and RPEs (including the resistance distance and the recently introduced stable and expressive positional encoding (SPE)) and compare their distinguishing power in terms of transformers. We believe our work will help navigate the huge number of choices of positional encoding and will provide guidance on the future design of positional encodings for graph transformers.


Getting lost in space: Large sample analysis of the commute distance

Neural Information Processing Systems

The commute distance between two vertices in a graph is the expected time it takes a random walk to travel from the first to the second vertex and back. We study the behavior of the commute distance as the size of the underlying graph increases. We prove that the commute distance converges to an expression that does not take into account the structure of the graph at all and that is completely meaningless as a distance function on the graph. Consequently, the use of the raw commute distance for machine learning purposes is strongly discouraged for large graphs and in high dimensions. As an alternative we introduce the amplified commute distance that corrects for the undesired large sample effects.



Optimal transport distances for directed, weighted graphs: a case study with cell-cell communication networks

Nagai, James S., Costa, Ivan G., Schaub, Michael T.

arXiv.org Artificial Intelligence

Yet, extending OT-based distances to directed graphs is not simple, as a Comparing graphs by means of optimal transport has recently symmetric distance metric, typically derived from the distances between gained significant attention, as the distances induced by optimal nodes in the graph, is required within the cost function(s) typically transport provide both a principled metric between graphs as well employed within OT. To address this problem, here we consider as an interpretable description of the associated changes between two node-to-node distances, which have been developed for directed graphs in terms of a transport plan. As the lack of symmetry introduces graphs, namely, the Generalized Effective Resistance (GRD) [6] and challenges in the typically considered formulations, optimal Markov chain hitting time (HTD) [7]. Employing these distance transport distances for graphs have mostly been developed for undirected measures enables us to compute OT-based graph distances even for graphs. Here, we propose two distance measures to compare directed graphs. Specifically, we explore the use of these metrics for directed graphs based on variants of optimal transport: (i) an earth both Wasserstein (Earth Mover) and Gromov-Wasserstein based OT movers distance (Wasserstein) and (ii) a Gromov-Wasserstein (GW) distances for graphs [5].